Bartholomew County
Appendix of ' Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation '
We calculate it for each sequence x and average over the whole corpus. When decoding auto-regressively, the probabilities of the repetitive sentence loops also have a self-reinforcement effect. As shown in Figure 2, the probability of the token'located' increases almost The work was conducted in Apple. Here we use the end token to split sentences for ease of experiments. We present the probability of the token'located' ( y-axis) as the number of historical repetitions Best viewed in color and zoomed in a desktop monitor.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > Missouri > Buchanan County > Saint Joseph (0.04)
- (6 more...)
- Media > Film (1.00)
- Government (1.00)
- Leisure & Entertainment > Sports > Basketball (0.46)
Controlling What You Share: Assessing Language Model Adherence to Privacy Preferences
Ramírez, Guillem, Birch, Alexandra, Titov, Ivan
Large language models (LLMs) are primarily accessed via commercial APIs, but this often requires users to expose their data to service providers. In this paper, we explore how users can stay in control of their data by using privacy profiles: simple natural language instructions that say what should and should not be revealed. We build a framework where a local model uses these instructions to rewrite queries, only hiding details deemed sensitive by the user, before sending them to an external model, thus balancing privacy with performance. To support this research, we introduce PEEP, a multilingual dataset of real user queries annotated to mark private content and paired with synthetic privacy profiles. Experiments with lightweight local LLMs show that, after fine-tuning, they not only achieve markedly better privacy preservation but also match or exceed the performance of much larger zero-shot models. At the same time, the system still faces challenges in fully adhering to user instructions, underscoring the need for models with a better understanding of user-defined privacy preferences.
- Europe > Austria > Vienna (0.14)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- North America > United States > Indiana > Bartholomew County > Columbus (0.04)
- (7 more...)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (1.00)
Appendix of ' Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation '
We calculate it for each sequence x and average over the whole corpus. When decoding auto-regressively, the probabilities of the repetitive sentence loops also have a self-reinforcement effect. As shown in Figure 2, the probability of the token'located' increases almost The work was conducted in Apple. Here we use the end token to split sentences for ease of experiments. We present the probability of the token'located' ( y-axis) as the number of historical repetitions Best viewed in color and zoomed in a desktop monitor.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > Missouri > Buchanan County > Saint Joseph (0.04)
- (6 more...)
- Media > Film (1.00)
- Government (1.00)
- Leisure & Entertainment > Sports > Basketball (0.46)
"A Big Bold Beautiful Journey" Is None of Those Things
"A Big Bold Beautiful Journey" Is None of Those Things Kogonada's fantasy film, starring Colin Farrell and Margot Robbie, suggests that a great directorial talent is losing his way. In Kogonada's new film, Colin Farrell and Margot Robbie try gamely to overcome the thinness with which their characters have been imagined. If movies were given scores as figure skaters are, fantasy would start with a high rating for technical difficulty. The landings of the genre are hard to stick, because fantasy, by definition, isn't rooted in experience. No one has lived on a distant planet, in the far future, or any place where dragons or wizards rule--so, kudos to anyone who can make such realms feel truly lived in.
- North America > United States > New York (0.05)
- North America > United States > Indiana > Bartholomew County > Columbus (0.04)
- North America > United States > California (0.04)
- Europe > Poland > Masovia Province > Warsaw (0.04)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
15,500 Seconds: Lean UAV Classification Using EfficientNet and Lightweight Fine-Tuning
Berg, Andrew P., Zhang, Qian, Wang, Mia Y.
As unmanned aerial vehicles (UAVs) become increasingly prevalent in both consumer and defense applications, the need for reliable, modality-specific classification systems grows in urgency. This paper addresses the challenge of data scarcity in UAV audio classification by expanding on prior work through the integration of pre-trained deep learning models, parameter-efficient fine-tuning (PEFT) strategies, and targeted data augmentation techniques. Using a custom dataset of 3,100 UAV audio clips (15,500 seconds) spanning 31 distinct drone types, we evaluate the performance of transformer-based and convolutional neural network (CNN) architectures under various fine-tuning configurations. Experiments were conducted with five-fold cross-validation, assessing accuracy, training efficiency, and robustness. Results show that full fine-tuning of the EfficientNet-B0 model with three augmentations achieved the highest validation accuracy (95.95), outperforming both the custom CNN and transformer-based models like AST. These findings suggest that combining lightweight architectures with PEFT and well-chosen augmentations provides an effective strategy for UAV audio classification on limited datasets. Future work will extend this framework to multimodal UAV classification using visual and radar telemetry.
- North America > United States > Indiana > Wayne County > Richmond (0.16)
- North America > United States > South Carolina > Charleston County > Charleston (0.15)
- North America > United States > Indiana > Tippecanoe County > West Lafayette (0.15)
- (5 more...)
A Hierarchical Test Platform for Vision Language Model (VLM)-Integrated Real-World Autonomous Driving
Zhou, Yupeng, Cui, Can, Peng, Juntong, Yang, Zichong, Lu, Juanwu, Panchal, Jitesh H, Yao, Bin, Wang, Ziran
Vision-Language Models (VLMs) have demonstrated notable promise in autonomous driving by offering the potential for multimodal reasoning through pretraining on extensive image-text pairs. However, adapting these models from broad web-scale data to the safety-critical context of driving presents a significant challenge, commonly referred to as domain shift. Existing simulation-based and dataset-driven evaluation methods, although valuable, often fail to capture the full complexity of real-world scenarios and cannot easily accommodate repeatable closed-loop testing with flexible scenario manipulation. In this paper, we introduce a hierarchical real-world test platform specifically designed to evaluate VLM-integrated autonomous driving systems. Our approach includes a modular, low-latency on-vehicle middleware that allows seamless incorporation of various VLMs, a clearly separated perception-planning-control architecture that can accommodate both VLM-based and conventional modules, and a configurable suite of real-world testing scenarios on a closed track that facilitates controlled yet authentic evaluations. We demonstrate the effectiveness of the proposed platform`s testing and evaluation ability with a case study involving a VLM-enabled autonomous vehicle, highlighting how our test framework supports robust experimentation under diverse conditions.
- North America > United States > Indiana > Tippecanoe County > West Lafayette (0.14)
- North America > United States > Indiana > Tippecanoe County > Lafayette (0.14)
- South America > Brazil > Rio de Janeiro > Rio de Janeiro (0.04)
- (14 more...)
- Transportation > Ground > Road (1.00)
- Automobiles & Trucks (1.00)
- Transportation > Infrastructure & Services (0.93)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
Efficient Multi-Agent Collaboration with Tool Use for Online Planning in Complex Table Question Answering
Zhou, Wei, Mesgar, Mohsen, Friedrich, Annemarie, Adel, Heike
Complex table question answering (TQA) aims to answer questions that require complex reasoning, such as multi-step or multi-category reasoning, over data represented in tabular form. Previous approaches demonstrated notable performance by leveraging either closed-source large language models (LLMs) or fine-tuned open-weight LLMs. However, fine-tuning LLMs requires high-quality training data, which is costly to obtain, and utilizing closed-source LLMs poses accessibility challenges and leads to reproducibility issues. In this paper, we propose Multi-Agent Collaboration with Tool use (MACT), a framework that requires neither closed-source models nor fine-tuning. In MACT, a planning agent and a coding agent that also make use of tools collaborate to answer questions. Our experiments on four TQA benchmarks show that MACT outperforms previous SoTA systems on three out of four benchmarks and that it performs comparably to the larger and more expensive closed-source model GPT-4 on two benchmarks, even when using only open-weight models without any fine-tuning. We conduct extensive analyses to prove the effectiveness of MACT's multi-agent collaboration in TQA.
- North America > Canada > Saskatchewan > Saskatoon (0.14)
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > Canada > Ontario > Toronto (0.14)
- (26 more...)
- Research Report (1.00)
- Financial News (0.68)
- Transportation > Passenger (1.00)
- Leisure & Entertainment > Sports > Soccer (1.00)
- Transportation > Air (0.93)
- Consumer Products & Services > Travel (0.93)
What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective
Li, Ming, Li, Yanhong, Zhou, Tianyi
What makes a difference in the post-training of LLMs? We investigate the training patterns of different layers in large language models (LLMs), through the lens of gradient, when training with different responses and initial models. We are specifically interested in how fast vs. slow thinking affects the layer-wise gradients, given the recent popularity of training LLMs on reasoning paths such as chain-of-thoughts (CoT) and process rewards. In our study, fast thinking without CoT leads to larger gradients and larger differences of gradients across layers than slow thinking (Detailed CoT), indicating the learning stability brought by the latter. Moreover, pre-trained LLMs are less affected by the instability of fast thinking than instruction-tuned LLMs. Additionally, we study whether the gradient patterns can reflect the correctness of responses when training different LLMs using slow vs. fast thinking paths. The results show that the gradients of slow thinking can distinguish correct and irrelevant reasoning paths. As a comparison, we conduct similar gradient analyses on non-reasoning knowledge learning tasks, on which, however, trivially increasing the response length does not lead to similar behaviors of slow thinking. Our study strengthens fundamental understandings of LLM training and sheds novel insights on its efficiency and stability, which pave the way towards building a generalizable System-2 agent. Our code, data, and gradient statistics can be found in: https://github.com/MingLiiii/Layer_Gradient.
- Europe > United Kingdom > England > Greater London > London (0.14)
- Europe > United Kingdom > Scotland (0.04)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- (14 more...)
- Government > Regional Government > Europe Government > United Kingdom Government (0.92)
- Education (0.92)
- Transportation > Ground (0.92)
- (4 more...)
Large Language Models for Autonomous Driving (LLM4AD): Concept, Benchmark, Simulation, and Real-Vehicle Experiment
Cui, Can, Ma, Yunsheng, Yang, Zichong, Zhou, Yupeng, Liu, Peiran, Lu, Juanwu, Li, Lingxi, Chen, Yaobin, Panchal, Jitesh H., Abdelraouf, Amr, Gupta, Rohit, Han, Kyungtae, Wang, Ziran
With the broader usage and highly successful development of Large Language Models (LLMs), there has been a growth of interest and demand for applying LLMs to autonomous driving technology. Driven by their natural language understanding and reasoning ability, LLMs have the potential to enhance various aspects of autonomous driving systems, from perception and scene understanding to language interaction and decision-making. In this paper, we first introduce novel concepts and approaches to designing LLMs for autonomous driving (LLM4AD). Then, we propose a comprehensive benchmark for evaluating the instruction-following abilities of LLMs within the autonomous driving domain. Furthermore, we conduct a series of experiments on both simulation and real-world vehicle platforms, thoroughly evaluating the performance and potential of our LLM4AD systems. Our research highlights the significant potential of LLMs to enhance various aspects of autonomous vehicle technology, from perception and scene understanding to language interaction and decision-making.
- North America > United States > Indiana > Tippecanoe County > West Lafayette (0.04)
- North America > United States > Indiana > Tippecanoe County > Lafayette (0.04)
- North America > United States > Texas > Travis County > Austin (0.04)
- (8 more...)
- Transportation > Ground > Road (1.00)
- Information Technology > Robotics & Automation (1.00)
- Automobiles & Trucks (1.00)
Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning
Yue, Yuanhao, Wang, Chengyu, Huang, Jun, Wang, Peng
The process of instruction tuning aligns pre-trained large language models (LLMs) with open-domain instructions and human-preferred responses. While several studies have explored autonomous approaches to distilling and annotating instructions from more powerful proprietary LLMs, such as ChatGPT, they often neglect the impact of task distributions and the varying difficulty of instructions of the training sets. This oversight can lead to imbalanced knowledge capabilities and poor generalization powers of small student LLMs. To address this challenge, we introduce Task-Aware Curriculum Planning for Instruction Refinement (TAPIR), a multi-round distillation framework with balanced task distributions and dynamic difficulty adjustment. This approach utilizes an oracle LLM to select instructions that are difficult for a student LLM to follow and distill instructions with balanced task distributions. By incorporating curriculum planning, our approach systematically escalates the difficulty levels, progressively enhancing the student LLM's capabilities. We rigorously evaluate TAPIR using two widely recognized benchmarks, including AlpacaEval 2.0 and MT-Bench. The empirical results demonstrate that the student LLMs, trained with our method and less training data, outperform larger instruction-tuned models and strong distillation baselines. The improvement is particularly notable in complex tasks, such as logical reasoning and code generation.
- North America > Canada > Ontario > Toronto (0.04)
- South America (0.04)
- Oceania (0.04)
- (10 more...)
- Government (1.00)
- Banking & Finance > Economy (1.00)
- Media > Music (0.68)
- Leisure & Entertainment > Sports > Football (0.46)